深度神经网络(DNNS)和数据集的增长不断上升,这激发了对同时选择和培训的有效解决方案的需求。许多迭代学习者的高参数优化方法(HPO)的许多方法,包括DNNS试图通过查询和学习响应表面来解决该问题的最佳表面来解决此问题。但是,这些方法中的许多方法都会产生近视疑问,不考虑有关响应结构的先验知识和/或执行偏见的成本感知搜索,当指定总成本预算时,所有这些都会加剧识别表现最好的模型。本文提出了一种新颖的方法,称为迭代学习者(BAPI),以在成本预算有限的情况下解决HPO问题。 BAPI是一种有效的非洋流贝叶斯优化解决方案,可以说明预算,并利用有关目标功能和成本功能的先验知识来选择更好的配置,并在评估期间(培训)做出更明智的决策。针对迭代学习者的不同HPO基准测试的实验表明,在大多数情况下,BAPI的性能比最先进的基线表现更好。
translated by 谷歌翻译
Previous work has shown the potential of deep learning to predict renal obstruction using kidney ultrasound images. However, these image-based classifiers have been trained with the goal of single-visit inference in mind. We compare methods from video action recognition (i.e. convolutional pooling, LSTM, TSM) to adapt single-visit convolutional models to handle multiple visit inference. We demonstrate that incorporating images from a patient's past hospital visits provides only a small benefit for the prediction of obstructive hydronephrosis. Therefore, inclusion of prior ultrasounds is beneficial, but prediction based on the latest ultrasound is sufficient for patient risk stratification.
translated by 谷歌翻译
Abusive language is a concerning problem in online social media. Past research on detecting abusive language covers different platforms, languages, demographies, etc. However, models trained using these datasets do not perform well in cross-domain evaluation settings. To overcome this, a common strategy is to use a few samples from the target domain to train models to get better performance in that domain (cross-domain few-shot training). However, this might cause the models to overfit the artefacts of those samples. A compelling solution could be to guide the models toward rationales, i.e., spans of text that justify the text's label. This method has been found to improve model performance in the in-domain setting across various NLP tasks. In this paper, we propose RAFT (Rationale Adaptor for Few-shoT classification) for abusive language detection. We first build a multitask learning setup to jointly learn rationales, targets, and labels, and find a significant improvement of 6% macro F1 on the rationale detection task over training solely rationale classifiers. We introduce two rationale-integrated BERT-based architectures (the RAFT models) and evaluate our systems over five different abusive language datasets, finding that in the few-shot classification setting, RAFT-based models outperform baseline models by about 7% in macro F1 scores and perform competitively to models finetuned on other source domains. Furthermore, RAFT-based models outperform LIME/SHAP-based approaches in terms of plausibility and are close in performance in terms of faithfulness.
translated by 谷歌翻译
Humans have perfected the art of learning from multiple modalities through sensory organs. Despite their impressive predictive performance on a single modality, neural networks cannot reach human level accuracy with respect to multiple modalities. This is a particularly challenging task due to variations in the structure of respective modalities. Conditional Batch Normalization (CBN) is a popular method that was proposed to learn contextual features to aid deep learning tasks. This technique uses auxiliary data to improve representational power by learning affine transformations for convolutional neural networks. Despite the boost in performance observed by using CBN layers, our work reveals that the visual features learned by introducing auxiliary data via CBN deteriorates. We perform comprehensive experiments to evaluate the brittleness of CBN networks to various datasets, suggesting that learning from visual features alone could often be superior for generalization. We evaluate CBN models on natural images for bird classification and histology images for cancer type classification. We observe that the CBN network learns close to no visual features on the bird classification dataset and partial visual features on the histology dataset. Our extensive experiments reveal that CBN may encourage shortcut learning between the auxiliary data and labels.
translated by 谷歌翻译
Machine learning models rely on various assumptions to attain high accuracy. One of the preliminary assumptions of these models is the independent and identical distribution, which suggests that the train and test data are sampled from the same distribution. However, this assumption seldom holds in the real world due to distribution shifts. As a result models that rely on this assumption exhibit poor generalization capabilities. Over the recent years, dedicated efforts have been made to improve the generalization capabilities of these models collectively known as -- \textit{domain generalization methods}. The primary idea behind these methods is to identify stable features or mechanisms that remain invariant across the different distributions. Many generalization approaches employ causal theories to describe invariance since causality and invariance are inextricably intertwined. However, current surveys deal with the causality-aware domain generalization methods on a very high-level. Furthermore, we argue that it is possible to categorize the methods based on how causality is leveraged in that method and in which part of the model pipeline is it used. To this end, we categorize the causal domain generalization methods into three categories, namely, (i) Invariance via Causal Data Augmentation methods which are applied during the data pre-processing stage, (ii) Invariance via Causal representation learning methods that are utilized during the representation learning stage, and (iii) Invariance via Transferring Causal mechanisms methods that are applied during the classification stage of the pipeline. Furthermore, this survey includes in-depth insights into benchmark datasets and code repositories for domain generalization methods. We conclude the survey with insights and discussions on future directions.
translated by 谷歌翻译
在本文中,我们介绍了计算机视觉研讨会上的女性 - WICV 2022,与路易斯安那州新奥尔良的混合CVPR 2022一起组织。它为计算机视觉社区中的少数(女性)群体提供了声音,并着重于提高这些研究人员在学术界和工业中的可见性。 WICV认为,这样的事件可以在降低计算机视觉领域的性别失衡方面发挥重要作用。 WICV每年都会组织a)a)从少数群体的研究人员之间合作的机会,b)指导女性初级研究人员,c)向演示者提供财政支持,以克服货币负担,D)榜样的大量选择,他们可以在职业生涯开始时,是年轻研究人员的例子。在本文中,我们介绍了有关研讨会计划的报告,过去几年的趋势,关于WICV 2022讲习班的演示者,与会者和赞助的统计摘要。
translated by 谷歌翻译
在线社交网络(OSN)有助于访问各种数据,使研究人员能够分析用户的行为并开发用户行为分析模型。这些模型在很大程度上依赖于观察到的数据,这些数据通常由于参与不平等而产生偏差。这种不平等由三组在线用户组成:潜伏者 - 仅消耗内容的用户,招聘者 - 对内容创建的用户和贡献者很少贡献 - 负责创建大多数在线内容的用户。在解释人口水平的利益或情感的同时,未能考虑所有群体的贡献,可能会产生偏见的结果。为了减少贡献者引起的偏见,在这项工作中,我们专注于强调参与者在观察到的数据中的贡献,因为与潜伏者相比,它们更有可能贡献,与贡献者相比,它们的人口更大。这些用户行为分析的第一步是找到他们接触但没有互动的主题。为此,我们提出了一个新颖的框架,有助于识别这些用户并估算其主题曝光。暴露估计机制是通过合并来自类似贡献者的行为模式以及用户的人口统计学和个人资料信息来建模的。
translated by 谷歌翻译
不依赖虚假相关性的学习预测因素涉及建立因果关系。但是,学习这样的表示非常具有挑战性。因此,我们制定了从高维数据中学习因果表示的问题,并通过合成数据研究因果恢复。这项工作引入了贝叶斯因果发现的潜在变量解码器模型BCD,并在轻度监督和无监督的环境中进行实验。我们提出了一系列合成实验,以表征因果发现的重要因素,并表明将已知的干预靶标用作标签有助于无监督的贝叶斯推断,对线性高斯添加噪声潜在结构性因果模型的结构和参数。
translated by 谷歌翻译
尽管Covid-19疫苗对病毒取得了惊人的成功,但很大一部分人口仍然不愿接受疫苗接种,这破坏了政府控制该病毒的努力。为了解决这个问题,我们需要了解导致这种行为的不同因素,包括社交媒体话语,新闻媒体宣传,政府的回应,人口统计和社会经济地位以及COVID-19统计等等。涵盖所有这些方面,使得在推断疫苗犹豫的问题时很难形成完整的情况。在本文中,我们构建了一个多源,多模式和多功能在线数据存储库Covaxnet。我们提供描述性分析和见解,以说明Covaxnet中的关键模式。此外,我们提出了一种新颖的方法来连接在线和离线数据,以促进利用互补信息源的推理任务。
translated by 谷歌翻译
知识图是捕获实体和实体之间关系的语义知识的绝佳资源。但是,当前的深度学习模型将其作为输入分布式表示或向量。因此,该图在矢量化表示中被压缩。我们进行了一项研究,以检查深度学习模型是否可以压缩图形,然后以大多数语义完整输出相同的图。我们的实验表明,变压器模型无法表达输入知识图的完整语义。我们发现这是由于知识图中包含的定向,关系和基于类型的信息与变压器注意矩阵的完全连接的令牌无向图形解释之间的差异。
translated by 谷歌翻译